Wider and Deeper, Cheaper and Faster: Tensorized LSTMs for Sequence Learning
نویسندگان
چکیده
Long Short-Term Memory (LSTM) is a popular approach to boosting the ability of Recurrent Neural Networks to store longer term temporal information. The capacity of an LSTM network can be increased by widening and adding layers. However, usually the former introduces additional parameters, while the latter increases the runtime. As an alternative we propose the Tensorized LSTM in which the hidden states are represented by tensors and updated via a cross-layer convolution. By increasing the tensor size, the network can be widened efficiently without additional parameters since the parameters are shared across different locations in the tensor; by delaying the output, the network can be deepened implicitly with little additional runtime since deep computations for each timestep are merged into temporal computations of the sequence. Experiments conducted on five challenging sequence learning tasks show the potential of the proposed model.
منابع مشابه
ICT Mediated Learning for Socio-Economic Empowerment
ICT can advance the learning process by making it faster, cheaper, and wider reaching that was not possible before. This form of learning can be treated as an interactive one among many and supporting the improvement of this process is expected to produce better result. However, innovative processes have to be incorporated both in terms of pedagogy and technology. Pedagogy should be universal a...
متن کاملFast and Accurate Sequence Labeling with Iterated Dilated Convolutions
Today when many practitioners run basic NLP on the entire web and large-volume traffic, faster methods are paramount to saving time and energy costs. Recent advances in GPU hardware have led to the emergence of bi-directional LSTMs as a standard method for obtaining pertoken vector representations serving as input to labeling tasks such as NER (often followed by prediction in a linear-chain CRF...
متن کاملDeep Fusion LSTMs for Text Semantic Matching
Recently, there is rising interest in modelling the interactions of text pair with deep neural networks. In this paper, we propose a model of deep fusion LSTMs (DF-LSTMs) to model the strong interaction of text pair in a recursive matching way. Specifically, DF-LSTMs consist of two interdependent LSTMs, each of which models a sequence under the influence of another. We also use external memory ...
متن کاملUnsupervised Learning of Video Representations using LSTMs
We use multilayer Long Short Term Memory (LSTM) networks to learn representations of video sequences. Our model uses an encoder LSTM to map an input sequence into a fixed length representation. This representation is decoded using single or multiple decoder LSTMs to perform different tasks, such as reconstructing the input sequence, or predicting the future sequence. We experiment with two kind...
متن کاملمقایسه قابلیت یادگیری حرکتی سکانسی در افراد سالمند و میانسال
Introduction: Acquisition motor skill training in older adults are great importance. The main purpose of this study was to compare middle age and older adult’s implicit learning in serial reaction time task. Methods: A serial reaction time task by using software was applied for studying motor learning in 30 older and 30 middle age adults. In this task 4 squares with different colors appeared ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2017